232 research outputs found
A Measurement Framework for Directed Networks
Partially-observed network data collected by link-tracing based sampling
methods is often being studied to obtain the characteristics of a large complex
network. However, little attention has been paid to sampling from directed
networks such as WWW and Peer-to-Peer networks. In this paper, we propose a
novel two-step (sampling/estimation) framework to measure nodal characteristics
which can be defined by an average target function in an arbitrary directed
network. To this end, we propose a personalized PageRank-based algorithm to
visit and sample nodes. This algorithm only uses already visited nodes as local
information without any prior knowledge about the latent structure of the
network. Moreover, we introduce a new estimator based on the approximate
importance sampling to estimate average target functions. The proposed
estimator utilizes calculated PageRank value of each sampled node as an
approximation for the exact visiting probability. To the best of our knowledge,
this is the first study on correcting the bias of a sampling method by
re-weighting of measured values that considers the effect of approximation of
visiting probabilities. Comprehensive theoretical and empirical analysis of the
estimator demonstrate that it is asymptotically unbiased even in situations
where stationary distribution of PageRank is poorly approximated.Comment: 10 pages, 6 figure
Performance-Optimum Superscalar Architecture for Embedded Applications
Embedded applications are widely used in portable devices such as wireless
phones, personal digital assistants, laptops, etc. High throughput and real
time requirements are especially important in such data-intensive tasks.
Therefore, architectures that provide the required performance are the most
desirable. On the other hand, processor performance is severely related to the
average memory access delay, number of processor registers and also size of the
instruction window and superscalar parameters. Therefore, cache, register file
and superscalar parameters are the major architectural concerns in designing a
superscalar architecture for embedded processors. Although increasing cache and
register file size leads to performance improvements in high performance
embedded processors, the increased area, power consumption and memory delay are
the overheads of these techniques. This paper explores the effect of cache,
register file and superscalar parameters on the processor performance to
specify the optimum size of these parameters for embedded applications.
Experimental results show that although having bigger size of these parameters
is one of the performance improvement approaches in embedded processors,
however, by increasing the size of some parameters over a threshold value,
performance improvement is saturated and especially in cache size, increments
over this threshold value decrease the performance
Boundary-bulk interplay of molecular motor traffic flow through a compartment
The flow of motor proteins on a filamental track is modelled within the the
framework of lattice driven diffusive systems. Motors, considered as hopping
particles, perform a highly biased asymmetric exclusion process when bound to
the filament. With a certain rate, they detach from the filament and execute
unbiased random walk in the bulk which is considered as a closed cubic
compartment. Motors are injected (extracted) from the leftmost (rightmost) site
of the filament located along the symmetry axis of the compartment. We explore
the transport properties of this system and investigate the bulk-boundary
interplay on the system stationary states. It is shown that the detachment rate
notably affects the system properties. In particular and in contrast to ASEP,
it is shown that the density profile of bound particles exhibit different types
of non monotonic behaviours when the detachment rate varies. It is shown that
in certain situations, the density profile of the filament consists of
coexisting high and low regions.Comment: 9 pages, 17 eps figures, Revte
Diffusion of Innovations over Multiplex Social Networks
The ways in which an innovation (e.g., new behaviour, idea, technology,
product) diffuses among people can determine its success or failure. In this
paper, we address the problem of diffusion of innovations over multiplex social
networks where the neighbours of a person belong to one or multiple networks
(or layers) such as friends, families, or colleagues. To this end, we
generalise one of the basic game-theoretic diffusion models, called networked
coordination game, for multiplex networks. We present analytical results for
this extended model and validate them through a simulation study, finding among
other properties a lower bound for the success of an innovation.While simple
and leading to intuitively understandable results, to the best of our knowledge
this is the first extension of a game-theoretic innovation diffusion model for
multiplex networks and as such it provides a basic framework to study more
sophisticated innovation dynamics
Diffusion-Aware Sampling and Estimation in Information Diffusion Networks
Partially-observed data collected by sampling methods is often being studied
to obtain the characteristics of information diffusion networks. However, these
methods usually do not consider the behavior of diffusion process. In this
paper, we propose a novel two-step (sampling/estimation) measurement framework
by utilizing the diffusion process characteristics. To this end, we propose a
link-tracing based sampling design which uses the infection times as local
information without any knowledge about the latent structure of diffusion
network. To correct the bias of sampled data, we introduce three estimators for
different categories; link-based, node-based, and cascade-based. To the best of
our knowledge, this is the first attempt to introduce a complete measurement
framework for diffusion networks. We also show that the estimator plays an
important role in correcting the bias of sampling from diffusion networks. Our
comprehensive empirical analysis over large synthetic and real datasets
demonstrates that in average, the proposed framework outperforms the common BFS
and RW sampling methods in terms of link-based characteristics by about 37% and
35%, respectively.Comment: 8 pages, 4 figures, Published in: International Confernece on Social
Computing 2012 (SocialCom12
QANet: Tensor Decomposition Approach for Query-based Anomaly Detection in Heterogeneous Information Networks
Complex networks have now become integral parts of modern information
infrastructures. This paper proposes a user-centric method for detecting
anomalies in heterogeneous information networks, in which nodes and/or edges
might be from different types. In the proposed anomaly detection method, users
interact directly with the system and anomalous entities can be detected
through queries. Our approach is based on tensor decomposition and clustering
methods. We also propose a network generation model to construct synthetic
heterogeneous information network to test the performance of the proposed
method. The proposed anomaly detection method is compared with state-of-the-art
methods in both synthetic and real-world networks. Experimental results show
that the proposed tensor-based method considerably outperforms the existing
anomaly detection methods
Multidimensional epidemic thresholds in diffusion processes over interdependent networks
Several systems can be modeled as sets of interdependent networks where each
network contains distinct nodes. Diffusion processes like the spreading of a
disease or the propagation of information constitute fundamental phenomena
occurring over such coupled networks. In this paper we propose a new concept of
multidimensional epidemic threshold characterizing diffusion processes over
interdependent networks, allowing different diffusion rates on the different
networks and arbitrary degree distributions. We analytically derive and
numerically illustrate the conditions for multilayer epidemics, i.e., the
appearance of a giant connected component spanning all the networks.
Furthermore, we study the evolution of infection density and diffusion dynamics
with extensive simulation experiments on synthetic and real networks
Sampling from Diffusion Networks
The diffusion phenomenon has a remarkable impact on Online Social Networks
(OSNs). Gathering diffusion data over these large networks encounters many
challenges which can be alleviated by adopting a suitable sampling approach.
The contributions of this paper is twofold. First we study the sampling
approaches over diffusion networks, and for the first time, classify these
approaches into two categories; (1) Structure-based Sampling (SBS), and (2)
Diffusion-based Sampling (DBS). The dependency of the former approach to
topological features of the network, and unavailability of real diffusion paths
in the latter, converts the problem of choosing an appropriate sampling
approach to a trade-off. Second, we formally define the diffusion network
sampling problem and propose a number of new diffusion-based characteristics to
evaluate introduced sampling approaches. Our experiments on large scale
synthetic and real datasets show that although DBS performs much better than
SBS in higher sampling rates (16% ~ 29% on average), their performances differ
about 7% in lower sampling rates. Therefore, in real large scale systems with
low sampling rate requirements, SBS would be a better choice according to its
lower time complexity in gathering data compared to DBS. Moreover, we show that
the introduced sampling approaches (SBS and DBS) play a more important role
than the graph exploration techniques such as Breadth-First Search (BFS) and
Random Walk (RW) in the analysis of diffusion processes.Comment: Published in Proceedings of the 2012 International Conference on
Social Informatics, Pages 106-11
NetSpam: a Network-based Spam Detection Framework for Reviews in Online Social Media
Nowadays, a big part of people rely on available content in social media in
their decisions (e.g. reviews and feedback on a topic or product). The
possibility that anybody can leave a review provide a golden opportunity for
spammers to write spam reviews about products and services for different
interests. Identifying these spammers and the spam content is a hot topic of
research and although a considerable number of studies have been done recently
toward this end, but so far the methodologies put forth still barely detect
spam reviews, and none of them show the importance of each extracted feature
type. In this study, we propose a novel framework, named NetSpam, which
utilizes spam features for modeling review datasets as heterogeneous
information networks to map spam detection procedure into a classification
problem in such networks. Using the importance of spam features help us to
obtain better results in terms of different metrics experimented on real-world
review datasets from Yelp and Amazon websites. The results show that NetSpam
outperforms the existing methods and among four categories of features;
including review-behavioral, user-behavioral, reviewlinguistic,
user-linguistic, the first type of features performs better than the other
categories
Link Prediction in Multiplex Networks based on Interlayer Similarity
Some networked systems can be better modelled by multilayer structure where
the individual nodes develop relationships in multiple layers. Multilayer
networks with similar nodes across layers are also known as multiplex networks.
This manuscript proposes a novel framework for predicting forthcoming or
missing links in multiplex networks. The link prediction problem in multiplex
networks is how to predict links in one of the layers, taking into account the
structural information of other layers. The proposed link prediction framework
is based on interlayer similarity and proximity-based features extracted from
the layer for which the link prediction is considered. To this end, commonly
used proximity-based features such as Adamic-Adar and Jaccard Coefficient are
considered. These features that have been originally proposed to predict
missing links in monolayer networks, do not require learning, and thus are
simple to compute. The proposed method introduces a systematic approach to take
into account interlayer similarity for the link prediction purpose.
Experimental results on both synthetic and real multiplex networks reveal the
effectiveness of the proposed method and show its superior performance than
state-of-the-art algorithms proposed for the link prediction problem in
multiplex networks
- …